Data-driven control of spatiotemporal chaos with reduced-order neural ODE-based models and reinforcement learning
نویسندگان
چکیده
Deep reinforcement learning (RL) is a data-driven method capable of discovering complex control strategies for high-dimensional systems, making it promising flow applications. In particular, the present work motivated by goal reducing energy dissipation in turbulent flows, and example considered spatiotemporally chaotic dynamics Kuramoto-Sivashinsky equation (KSE). A major challenge associated with RL that substantial training data must be generated repeatedly interacting target system, costly when system computationally or experimentally expensive. We mitigate this manner combining dimensionality reduction via an autoencoder neural ODE framework to obtain low-dimensional dynamical model from just limited set. substitute reduced-order (ROM) place true during efficiently estimate optimal policy, which can then deployed on system. For KSE actuated localized forcing ("jets") at four locations, we demonstrate are able learn ROM accurately captures as well underlying natural snapshots experiencing random actuations. Using objective minimizing power cost, extract policy using deep RL. show ROM-based strategy translates highlight agent discovers stabilizes forced equilibrium solution captured discovered through related existing known KSE.
منابع مشابه
Reduced-Order Models for Data-Limited Reinforcement Learning
Often in reinforcement learning we are confronted with planning in a domain with an unknown dynamics model. In poorly understood domains where an exact model of either the dynamics, value function, or policy is unavailable we commonly choose to use large models, consisting of many parameters. Unfortunately, these large models can require a prohibitively large amount of training in realworld dom...
متن کاملControlling chaos by GA-based reinforcement learning neural network
This paper proposes a TD (temporal difference) and GA (genetic algorithm) based reinforcement (TDGAR) neural learning scheme for controlling chaotic dynamical systems based on the technique of small perturbations. The TDGAR learning scheme is a new hybrid GA, which integrates the TD prediction method and the GA to fulfill the reinforcement learning task. Structurely, the TDGAR learning system i...
متن کاملReinforcement Learning Based PID Control of Wind Energy Conversion Systems
In this paper an adaptive PID controller for Wind Energy Conversion Systems (WECS) has been developed. Theadaptation technique applied to this controller is based on Reinforcement Learning (RL) theory. Nonlinearcharacteristics of wind variations as plant input, wind turbine structure and generator operational behaviordemand for high quality adaptive controller to ensure both robust stability an...
متن کاملDriven Heisenberg Magnets : Nonequilibrium Criticality , Spatiotemporal Chaos and Control
PACS. 64.60.Cn – Order-disorder transformations; statistical mechanics of model systems. Abstract. – We drive a d-dimensional Heisenberg magnet using an anisotropic current. The continuum Langevin equation is analysed using a dynamical renormalization group and numerical simulations. We discover a rich steady-state phase diagram, including a critical point in a new nonequilibrium universality c...
متن کاملDelay driven spatiotemporal chaos in single species population dynamics models.
Questions surrounding the prevalence of complex population dynamics form one of the central themes in ecology. Limit cycles and spatiotemporal chaos are examples that have been widely recognised theoretically, although their importance and applicability to natural populations remains debatable. The ecological processes underlying such dynamics are thought to be numerous, though there seems to b...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Proceedings of The Royal Society A: Mathematical, Physical and Engineering Sciences
سال: 2022
ISSN: ['1471-2946', '1364-5021']
DOI: https://doi.org/10.1098/rspa.2022.0297